Synthesizing multimodal utterances for conversational agents
نویسندگان
چکیده
Conversational agents are supposed to combine speech with non-verbal modalities for intelligible multimodal utterances. In this paper, we focus on the generation of gesture and speech fromXML-based descriptions of their overt form. An incremental production model is presented that combines the synthesis of synchronized gestural, verbal, and facial behaviors with mechanisms for linking them in fluent utterances with natural co-articulation and transition effects. In particular, an efficient kinematic approach for animating hand gestures from shape specifications is presented, which provides fine adaptation to temporal constraints that are imposed by cross-modal synchrony. Copyright # 2004 John Wiley & Sons, Ltd.
منابع مشابه
Dialogue with Agents in a Dynamic World
We have constructed the discourse component for a multimodal user-interface dialogue system. The larger system acts as a conversational agent engaged in dialogue with the human user. Output originating with the backend application system can take the eventual form of spoken English utterances, printed English text, or changes in the graphical display screen, e.g., appearance, disappearance, mov...
متن کاملMaking Relative Sense: From Word-Graphs To Semantic Frames
Scaling up from controlled single domain spoken dialogue systems towards conversational, multi-domain and multimodal dialogue systems poses new challenges for the reliable processing of less restricted user utterances. In this paper we explore the feasibility to employ a general purpose ontology for various tasks involved in processing the user’s utterances.
متن کاملModel-based Animation of Coverbal Gesture
Virtual conversational agents are supposed to combine speech with nonverbal modalities for intelligible and believeable utterances. However, the automatic synthesis of coverbal gestures still struggles with several problems like naturalness in procedurally generated animations, flexibility in pre-defined movements, and synchronization with speech. In this paper, we focus on generating complex m...
متن کاملAdaptive expressiveness: virtual conversational agents that can align to their interaction partner
Speakers in dialogue tend to adapt to each other by starting to use similar lexical items, syntactic structures, or gestures. This behaviour, called alignment, may serve important cognitive, communicative and social functions (such as speech facilitation, grounding and rapport). Our aim is to enable and study the effects of these subtle aspects of communication in virtual conversational agents....
متن کاملGeneration and evaluation of user tailored responses in multimodal dialogue
When people engage in conversation, they tailor their utterances to their conversational partners, whether these partners are other humans or computational systems. This tailoring, or adaptation to the partner takes place in all facets of human language use, and is based on amental model or a user model of the conversational partner. Such adaptation has been shown to improve listeners’ comprehe...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Journal of Visualization and Computer Animation
دوره 15 شماره
صفحات -
تاریخ انتشار 2004